17 research outputs found
Factorization of Discriminatively Trained i-vector Extractor for Speaker Recognition
In this work, we continue in our research on i-vector extractor for speaker
verification (SV) and we optimize its architecture for fast and effective
discriminative training. We were motivated by computational and memory
requirements caused by the large number of parameters of the original
generative i-vector model. Our aim is to preserve the power of the original
generative model, and at the same time focus the model towards extraction of
speaker-related information. We show that it is possible to represent a
standard generative i-vector extractor by a model with significantly less
parameters and obtain similar performance on SV tasks. We can further refine
this compact model by discriminative training and obtain i-vectors that lead to
better performance on various SV benchmarks representing different acoustic
domains.Comment: Submitted to Interspeech 2019, Graz, Austria. arXiv admin note:
substantial text overlap with arXiv:1810.1318
Exploiting i–vector posterior covariances for short–duration language recognition
Linear models in i-vector space have shown to be an effective solution not only for speaker identification, but also for language recogniton. The i-vector extraction process, however, is affected by several factors, such as noise level, the acoustic content of the utterance and the duration of the spoken segments. These factors influence both the i-vector estimate and its uncertainty, represented by the i-vector posterior covariance matrix. Modeling of i-vector uncertainty with Probabilistic Linear Discriminant Analysis has shown to be effective for short-duration speaker identification. This paper extends the approach to language recognition, analyzing the effects of i-vector covariances on a state-of-the-art Gaussian classifier, and proposes an effective solution for the reduction of the average detection cost (Cavg) for short segments
An attention-based backend allowing efficient fine-tuning of transformer models for speaker verification
In recent years, self-supervised learning paradigm has received extensive
attention due to its great success in various down-stream tasks. However, the
fine-tuning strategies for adapting those pre-trained models to speaker
verification task have yet to be fully explored. In this paper, we analyze
several feature extraction approaches built on top of a pre-trained model, as
well as regularization and learning rate schedule to stabilize the fine-tuning
process and further boost performance: multi-head factorized attentive pooling
is proposed to factorize the comparison of speaker representations into
multiple phonetic clusters. We regularize towards the parameters of the
pre-trained model and we set different learning rates for each layer of the
pre-trained model during fine-tuning. The experimental results show our method
can significantly shorten the training time to 4 hours and achieve SOTA
performance: 0.59%, 0.79% and 1.77% EER on Vox1-O, Vox1-E and Vox1-H,
respectively.Comment: Accepted by SLT202
Automatic language identification using deep neural networks
Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. I. LĂłpez-Moreno, J. González-DomĂnguez, P. Oldrich, D. R. MartĂnez, J. González-RodrĂguez, "Automatic language identification using deep neural networks", IEEE International Conference on Acoustics, Speech, and Signal Processing ICASSP, Florence (Italy), 2014This work studies the use of deep neural networks (DNNs)
to address automatic language identification (LID). Motivated
by their recent success in acoustic modelling, we adapt DNNs
to the problem of identifying the language of a given spoken
utterance from short-term acoustic features. The proposed approach
is compared to state-of-the-art i-vector based acoustic
systems on two different datasets: Google 5M LID corpus and
NIST LRE 2009. Results show how LID can largely benefit
from using DNNs, especially when a large amount of training
data is available. We found relative improvements up to 70%,
in Cavg, over the baseline system
On the Use of Deep Feedforward Neural Networks for Automatic Language Identification
In this work, we present a comprehensive study on the use of deep neural networks (DNNs) for automatic language identification (LID). Motivated by the recent success of using DNNs in acoustic modeling for speech recognition, we adapt DNNs to the problem of identifying the language in a given utterance from its short-term acoustic features. We propose two different DNN- based approaches. In the first one, the DNN acts as an end-to-end LID classifier, receiving as input the speech features and providing as output the estimated probabilities of the target languages. In the second approach, the DNN is used to extract bottleneck features that are then used as inputs for a state-of-the-art i-vector system. Experiments are conducted in two different scenarios: the complete NIST Language Recognition Evaluation dataset 2009 (LRE’09) and a subset of the Voice of America (VOA) data from LRE’09, in which all languages have the same amount of training data. Results for both datasets demonstrate that the DNN-based systems significantly outperform a state-of-art i-vector system when dealing with short-duration utterances. Furthermore, the combination of the DNN-based and the classical i-vector system leads to additional performance improvements (up to 45% of relative improvement in both EER and Cavg on 3s and 10s conditions, respectively)